---
title: AWS S3
section_name: Data
maturity: public-preview
---

# AWS S3 {: #aws-s3 }

!!! info "Public preview"
    The native S3 connector, which introduces performance improvements and additional feature support, is on by default.

    <b>Feature flag:</b> Enable S3 Connector

## Supported authentication {: #supported-authentication }

- [AWS credentials:](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_credentials_access-keys.html){ target=_blank } Long-term credentials using `aws_access_key_id` and `aws_secret_access_key`.
- [AWS temporary credentials:](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_credentials_temp.html){ target=_blank } Short-term credentials using `aws_access_key_id`, `aws_secret_access_key`, and `aws_session_token`.

## Prerequisites {: #prerequisites }

The following is required before connecting to AWS S3 in DataRobot:

- AWS S3 account

!!! note "Self-Managed AI Platform installations"
    For Self-Managed AI Platform installations, an admin must [add the native S3 connector](connector#self-managed-ai-platform-admins).

## Required parameters {: #required-parameters }

The table below lists the minimum required fields to establish a connection with AWS S3:

Required field | Description |  Documentation
--------------- | ---------- | -----------
`bucketName` |     A container that stores your data in AWS S3.        |          [AWS documentation](https://docs.aws.amazon.com/AmazonS3/latest/userguide/UsingBucket.html){ target=_blank }

Note that you can specify `bucketRegion` under **Show advanced options**, however this parameter is not required.

## Feature considerations {: #feature-considerations }

- This connector _does not_ support:

    - Feature Discovery projects
    - Data wrangling in Workbench
    - Adding dynamic datasets to a Workbench Use Case

- This connector _does_ support:

    - Creating and sharing AWS credentials using [secure configurations](secure-config).
    - Parquet file ingest including, single parquet files, partitioned parquet files in a folder, and zipped parquet files.



